The Performance of Group Lasso for Linear Regression of Grouped Variables
نویسندگان
چکیده
The lasso [19] and group lasso [23] are popular algorithms in the signal processing and statistics communities. In signal processing, these algorithms allow for efficient sparse approximations of arbitrary signals in overcomplete dictionaries. In statistics, they facilitate efficient variable selection and reliable regression under the linear model assumption. In both cases, there is now ample empirical evidence to suggest that an appropriately regularized group lasso can outperform the lasso whenever there is a natural grouping of the dictionary elements/regression variables in terms of their contributions to the observations [1, 23]. Our goal in this technical report is to analytically characterize the regression performance of the group lasso algorithm using `1/`2 regularization for the case in which one can have far more regression variables than observations. Analytical characterization of group lasso in this “underdetermined” setting has received some attention lately in the statistics literature [1, 14–17]. However, prior analytical work on the performance of group lasso either studies an asymptotic regime [1, 15–17], focuses on random design matrices [1, 16], and/or relies on metrics that are computationally expensive to evaluate [14, 15, 17]. Recently, Candés and Plan [4] successfully circumvented somewhat similar shortcomings
منابع مشابه
Regularization Path Algorithms for Detecting Gene Interactions
In this study, we consider several regularization path algorithms with grouped variable selection for modeling gene-interactions. When fitting with categorical factors, including the genotype measurements, we often define a set of dummy variables that represent a single factor/interaction of factors. Yuan & Lin (2006) proposed the groupLars and the group-Lasso methods through which these groups...
متن کاملBlockwise Sparse Regression
Yuan an Lin (2004) proposed the grouped LASSO, which achieves shrinkage and selection simultaneously, as LASSO does, but works on blocks of covariates. That is, the grouped LASSO provides a model where some blocks of regression coefficients are exactly zero. The grouped LASSO is useful when there are meaningful blocks of covariates such as polynomial regression and dummy variables from categori...
متن کاملDifferenced-Based Double Shrinking in Partial Linear Models
Partial linear model is very flexible when the relation between the covariates and responses, either parametric and nonparametric. However, estimation of the regression coefficients is challenging since one must also estimate the nonparametric component simultaneously. As a remedy, the differencing approach, to eliminate the nonparametric component and estimate the regression coefficients, can ...
متن کاملRobust Variable Selection in Functional Linear Models
We consider the problem of selecting functional variables using the L1 regularization in a functional linear regression model with a scalar response and functional predictors in the presence of outliers. Since the LASSO is a special case of the penalized least squares regression with L1-penalty function it suffers from the heavy-tailed errors and/or outliers in data. Recently, the LAD regressio...
متن کاملRegression Performance of Group Lasso for Arbitrary Design Matrices
In many linear regression problems, explanatory variables are activated in groups or clusters; group lasso has been proposed for regression in such cases. This paper studies the nonasymptotic regression performance of group lasso using `1/`2 regularization for arbitrary (random or deterministic) design matrices. In particular, the paper establishes under a statistical prior on the set of nonzer...
متن کامل